Accessibility settings

Published on in Vol 10 (2026)

This is a member publication of King's College London (Jisc)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/90139, first published .
Retrieval-Augmented Generation Versus GPT-4o for Patient-Facing Gynecological Cancer Information: Quality Evaluation

Retrieval-Augmented Generation Versus GPT-4o for Patient-Facing Gynecological Cancer Information: Quality Evaluation

Retrieval-Augmented Generation Versus GPT-4o for Patient-Facing Gynecological Cancer Information: Quality Evaluation

1Critical Care, University College London Hospitals NHS Foundation Trust, London, England, United Kingdom

2University College London Hospitals NHS Foundation Trust, London, United Kingdom

3Anaesthetic Department, Whittington Health NHS Trust, London, England, United Kingdom

4School of Nursing and Midwifery, London South Bank University, London, England, United Kingdom

5Florence Nightingale Faculty of Nursing, Midwifery and Palliative Care, King's College London, King's College London, Strand, London, England, United Kingdom

*all authors contributed equally

Corresponding Author:

Louise Rose, PhD




Health literacy, which is the ability to obtain, process and understand basic health information, varies widely. Furthermore, situational stress influences the ability to comprehend health information [1]. High-quality accessible information underpins shared decision-making. Yet people with cancer frequently report information overload and unmet needs at diagnosis and during treatment [2]. Producing timely, personalized, and readable health-related materials is human-resource intensive, yields static documents, and can result in duplicated effort and variable quality [3]. Large language models (LLMs) may offer a solution through rapid generation and adaptation of text. However, LLMs may hallucinate, cite unverifiable sources, and produce materials misaligned with literacy needs [4]. Retrieval augmented generation (RAG) may address these issues as it grounds LLM outputs in verified information as opposed to proprietary LLMs that rely on training data alone [5].

Our objective was to compare base GPT-40 with GPT-4o enhanced with a tailored gynecological cancer knowledgebase (RAG GPT-4o) in order to determine whether RAG improves the quality of AI-generated patient-facing information.


Study Design

We compared two GPT-4o configurations to answer frequently asked gynecological cancer questions compiled from patient queries and service information resources. For the RAG GPT-4o intervention arm, a retrieval layer supplied passages from a curated UK knowledge base restricted to Macmillan Cancer Support [6] and The Eve Appeal [7] resources. The Base GPT-4o control arm generated answers without retrieval of external sources. RAG GPT-4o ran via an application programming interface (API); Base GPT-4o ran via the web interface.

We compiled 17 questions and generated two responses per question (Base GPT-4o via ChatGPT web interface; RAG GPT-4o via API), producing 34 outputs. Prompts were held constant across configurations. Paired model outputs for each question formed the unit of comparison. Seven expert raters, each, evaluated eight paired outputs, while an LLM-judge evaluated all 17 paired outputs. Application Programming Interface generation settings were recorded for reproducibility; equivalent sampling parameters could not be fixed in the web interface.

Recruitment

We recruited a convenience sample of clinical nurse specialists (CNSs) via professional networks. Participants were presented with paired outputs in random order and blinded to allocation. Each participant evaluated the same eight preselected question-answer pairs, the maximum feasible within one hour based on pilot testing, using the Quality Analysis of Medical AI (QAMAI) tool (six domains rated on 5-point Likert scales) [8]. Readability was assessed with the Linguistic Features Toolkit (LFTK), including Flesch Reading Ease (FRE), Flesch Kincaid Grade Level (FKGL), and total words [9]. An LLM-judge, implemented with RAG GPT-4o, also rated all paired outputs using the QAMAI tool. Prompt templates and model configuration details, including API settings, are provided.

Analysis

Two-tailed paired t-tests were used to compare Base and RAG scores when the within pair differences were approximately normally distributed; Wilcoxon signed-rank tests otherwise. Inter-rater agreement was assessed with intraclass correlation coefficients (ICCs 3,k); internal consistency with Cronbach α.

Ethical Considerations

This study was approved by the King’s College London Minimal Risk Research Ethics Committee (MRSU 24/25 46882). Participants received a participant information sheet, provided written informed consent, and could withdraw at any time. Privacy and confidentiality were protected throughout. Participants received a £25 voucher for participation.


Seven UK-based CNSs participated in the study (median post-registration experience 20 years; IQR 5‐30). Data collection ran from March-May 2025. Participants rated quality of RAG GPT-4o answers higher than those generated by base GPT-4o with highest mean difference in the domain, Provision of Sources and References (Table 1). Inter-rater agreement for total QAMAI scores was moderate (ICC 0.65, 95% CI 0.31-0.86); internal consistency was good (Cronbach α=0.81). The LLM-judge rated the quality of RAG GPT-4o answers higher than those generated by base GPT-4o (P<.001), again driven by the QAMAI Provision of Sources and References domain (P=.01) (Figure 1).

Figure 1. Mean QAMAI (Quality Analysis of Medical Artificial Intelligence) item-level domain scores (1 to 5, higher scores indicate higher quality) for RAG GPT-4o (RAG) and Base GPT-4o (ChatGPT). Left panel (Expert ratings): domain means calculated from 7 expert raters who each scored 8 matched answer pairs. Right panel (LLM ratings): domain means calculated from a single LLM-judge that scored all 17 matched answer pairs once. LLM: large language model; RAG: retrieval augmented generation.
Table 1. Paired analysis of QAMAIa and readability metrics.
QAMAI domainRAGb GPT-4o,
mean (SD)
Base GPT-4o, mean (SD)Mean difference
(95% CI)
P value
Participants: 8 question-answer pairs evaluated
Accuracy4.14 (0.75)3.98 (0.75)0.16 (−0.05 to 0.37).11
Clarity4.16 (0.65)4.02 (0.77)0.14 (−0.01 to 0.30).06
Relevance4.11 (0.73)4.09 (0.67)0.02 (−0.26 to 0.23).78
Completeness3.88 (0.90)3.84 (0.93)0.04 (−0.16 to 0.23).58
Provision of Sources & References3.98 (0.94)1.80 (1.24)2.18 (0.92 to 3.44).03
Usefulness3.96 (0.69)3.80 (0.64)0.16 (−0.10 to 0.42).17
Total (max=30)24.23 (3.60)21.54 (3.46)2.69 (0.77 to 4.62).01c
LLM-Judge: 17 question-answer pairs evaluated
Accuracy4.82 (0.39)4.06 (0.24)0.76 (0.53 to 0.99)<.001
Clarity4.82 (0.39)4.94 (0.24)−0.12 (−0.35 to 0.11).16
Relevance5.00 (0)5.00 (0)NAgNAg
Completeness3.94 (0.24)4.00 (0)0.06 (−0.18 to 0.06).32
Provision of Sources & References1.71 (0.77)1.06 (0.24)0.65 (0.25 to 1.05).01
Usefulness4.94 (0.24)4.71 (0.47)0.23 (−0.03 to 0.49).10
Total (max=30)25.24 (1.35)23.76 (0.56)1.48 (0.76 to 2.20)<.001c
LFTKd Domain
17 question-answer pairs evaluated
FKGLe7.2 (1.85)6.2 (1.87)1.00 (–0.1 to 2.2).08c
FREf67.9 (10.98)75.2 (9.54)−8.2 (–12.9 to –1.9).006c
Total words370.8 (84.27)308.5 (67.82)62.3 (34.6 to 90.0)<.001c

aQAMAI: Quality Analysis of Medical Artificial Intelligence.

bRAG: retrieval augmented generation.

cpaired t test (otherwise Wilcoxon signed-rank)

dLFTK: linguistic features toolkit.

eFKGL: Flesch Kincaid Grade Level.

fFRE: Flesch Reading Ease.

gNot applicable.

RAG GPT-4o answers were longer (P<.001) and had a lower FRE score (P=.006), indicating more difficult-to-read text. FKGL did not differ significantly (Table 1).


Overall, experts and the LLM-judge rated both configurations highly for accuracy, clarity, relevance, and readability. RAG GPT-4o achieved higher overall QAMAI scores than Base GPT-4o, driven mainly by better provision of sources and references rather than perceived gains in accuracy or clarity. Although the LLM-judge identified differences in source attribution, it appeared to weigh provenance less than expert raters or had difficulty applying QAMAI in this domain, consistent with other studies [10]. RAG GPT-4o outputs were longer and less easy to read.

Although overall QAMAI scores were high, this does not eliminate the risk of clinically important failure in patient-facing cancer information. Errors may present as confident hallucinated claims, omission of safety critical details such as red flag symptoms or treatment risks, or overgeneralization that fails to address question intent. Provenance related problems are also possible, including outdated guidance and references that do not support the accompanying statements. Within this small sample, ratings did not suggest prominent safety related concerns, but larger evaluations are needed to characterize the frequency and impact of these types of error, and to determine whether high rubric scores and source attribution are sufficient safeguards for deployment without clinical oversight.

Improved provenance is expected with RAG, but the marginal benefit observed here must be weighed against the overhead of curating and maintaining a knowledge base, monitoring retrieval quality, and updating content as guidance changes. Although transparent sourcing may support perceived credibility, acceptability and trust in patient-facing information are also shaped by readability, tone, length, and cognitive load. Future work should evaluate whether optimized RAG configurations, and patient centered presentation formats improve usability, comprehension, acceptability, and perceived trust without increasing cognitive burden.

These findings suggest LLMs can generate patient-facing information that clinicians may judge as accurate, relevant, and readable, and could reduce the time required to draft materials. In this study, RAG improved overall QAMAI scores primarily through source attribution rather than perceived improvements in accuracy or clarity. Because RAG outputs were longer and less easy to read, any gains in transparency should be balanced against potential impacts on comprehensibility and health literacy.

Study limitations include a small convenience sample, expert ratings restricted to eight of 17 questions, use of a single LLM and UK knowledge base, and interface asymmetry between API and web access. Because equivalent sampling parameters could not be fixed in the web interface, this comparison should be interpreted as pragmatic and exploratory.

Funding

This study was supported by the National Institute for Health and Care Research (NIHR) Pre-doctoral Clinical Academic Fellowship. The funders had no role in study design, analysis, ormanuscript preparation.

Data Availability

The datasets generated or analyzed during this study are available in the GitHub repository [11].

Conflicts of Interest

None declared.

  1. Berkman ND, Davis TC, McCormack L. Health literacy: what is it? J Health Commun. 2010;15 Suppl 2:9-19. [CrossRef] [Medline]
  2. Covvey JR, Kamal KM, Gorse EE, et al. Barriers and facilitators to shared decision-making in oncology: a systematic review of the literature. Support Care Cancer. May 2019;27(5):1613-1637. [CrossRef]
  3. Papadakos J, Giannopoulos E, Forbes L, et al. Reinventing the wheel: the incidence and cost implication of duplication of effort in patient education materials development. Patient Educ Couns. Jun 2021;104(6):1398-1405. [CrossRef] [Medline]
  4. Ji Z, Lee N, Frieske R, et al. Survey of hallucination in natural language generation. ACM Comput Surv. Dec 31, 2023;55(12):1-38. [CrossRef]
  5. Gupta S, Ranjan R, Singh SN. A comprehensive survey of retrieval-augmented generation (RAG): evolution, current landscape and future directions. arXiv. Preprint posted online on Oct 3, 2024. [CrossRef]
  6. Macmillan Cancer Support. 2025. URL: https://www.macmillan.org.uk/ [Accessed 2025-12-21]
  7. The Eve Appeal. 2025. URL: https://eveappeal.org.uk/ [Accessed 2025-12-21]
  8. Vaira LA, Lechien JR, Abbate V, et al. Validation of the Quality Analysis of Medical Artificial Intelligence (QAMAI) tool: a new tool to assess the quality of health information provided by AI platforms. Eur Arch Otorhinolaryngol. Nov 2024;281(11):6123-6131. [CrossRef]
  9. Lee BW, Lee JHJ. Handcrafted features in computational linguistics. arXiv. Preprint posted online on May 25, 2023. [CrossRef]
  10. Thakur AS, Choudhary K, Ramayapally VS, Vaidyanathan S, Hupkes D. Judging the judges: evaluating alignment and vulnerabilities in LLMs-as-judges. arXiv. Preprint posted online on Aug 18, 2025. [CrossRef]
  11. CDE-research-group/RAG-for-patient-information. GitHub. URL: https://github.com/CDE-Research-Group/RAG-for-Patient-Information [Accessed 2026-04-14]


API: application programming interface
CNS: clinical nurse specialist
FKGL: Flesch Kincaid Grade Level
FRE: Flesch Reading Ease
ICC: intraclass correlation coefficient
LFTK: linguistic features toolkit
LLM: large language models
QAMAI: Quality Analysis of Medical Artificial Intelligence
RAG: retrieval augmented generation


Edited by Amy Bucher; submitted 22.Dec.2025; peer-reviewed by Kuan-Hsun Lin, Ujjwal Pasupulety; final revised version received 12.Mar.2026; accepted 12.Mar.2026; published 24.Apr.2026.

Copyright

© Stephen Pearson, Mimi Reyburn, Conor Foley, Alison Finch, Suzanne Bench, Timothy Bonnici, Louise Rose. Originally published in JMIR Formative Research (https://formative.jmir.org), 24.Apr.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.